我们建议在没有观察到的变量的情况下,提出基于订购的方法,用于学习结构方程模型(SEM)的最大祖先图(MAG),直到其Markov等效类(MEC)。文献中的现有基于订购的方法通过学习因果顺序(C-order)恢复图。我们提倡一个名为“可移动顺序”(R-rorder)的新颖订单,因为它们比结构学习的C端口有利。这是因为R-orders是适当定义的优化问题的最小化器,该问题可以准确解决(使用强化学习方法)或大约(使用爬山搜索)。此外,R-orders(与C-orders不同)在MEC中的所有图表中都是不变的,并将C-orders包括为子集。鉴于一组R-orders通常明显大于C-orders集,因此优化问题更容易找到R级而不是C级。我们评估了在现实世界和随机生成的网络上提出的方法的性能和可伸缩性。
translated by 谷歌翻译
因果鉴定是因果推理文献的核心,在该文献中提出了完整的算法来识别感兴趣的因果问题。这些算法的有效性取决于访问正确指定的因果结构的限制性假设。在这项工作中,我们研究了可获得因果结构概率模型的环境。具体而言,因果图中的边缘是分配的概率,例如,可能代表来自领域专家的信念程度。另外,关于边缘的不确定的可能反映了特定统计检验的置信度。在这种情况下自然出现的问题是:给定这样的概率图和感兴趣的特定因果效应,哪些具有最高合理性的子图是什么?我们表明回答这个问题减少了解决NP-HARD组合优化问题,我们称之为边缘ID问题。我们提出有效的算法来近似此问题,并评估我们针对现实世界网络和随机生成图的算法。
translated by 谷歌翻译
We study experiment design for unique identification of the causal graph of a system where the graph may contain cycles. The presence of cycles in the structure introduces major challenges for experiment design as, unlike acyclic graphs, learning the skeleton of causal graphs with cycles may not be possible from merely the observational distribution. Furthermore, intervening on a variable in such graphs does not necessarily lead to orienting all the edges incident to it. In this paper, we propose an experiment design approach that can learn both cyclic and acyclic graphs and hence, unifies the task of experiment design for both types of graphs. We provide a lower bound on the number of experiments required to guarantee the unique identification of the causal graph in the worst case, showing that the proposed approach is order-optimal in terms of the number of experiments up to an additive logarithmic term. Moreover, we extend our result to the setting where the size of each experiment is bounded by a constant. For this case, we show that our approach is optimal in terms of the size of the largest experiment required for uniquely identifying the causal graph in the worst case.
translated by 谷歌翻译
我们研究在有关系统的结构侧信息时学习一组变量的贝叶斯网络(BN)的问题。众所周知,学习一般BN的结构在计算上和统计上具有挑战性。然而,通常在许多应用中,关于底层结构的侧面信息可能会降低学习复杂性。在本文中,我们开发了一种基于递归约束的算法,其有效地将这些知识(即侧信息)纳入学习过程。特别地,我们研究了关于底层BN的两种类型的结构侧信息:(i)其集团数的上限是已知的,或者(ii)它是无菱形的。我们为学习算法提供理论保证,包括每个场景所需的最坏情况的测试数量。由于我们的工作,我们表明可以通过多项式复杂性学习有界树木宽度BNS。此外,我们评估了综合性和现实世界结构的算法的性能和可扩展性,并表明它们优于最先进的结构学习算法。
translated by 谷歌翻译
Finding and localizing the conceptual changes in two scenes in terms of the presence or removal of objects in two images belonging to the same scene at different times in special care applications is of great significance. This is mainly due to the fact that addition or removal of important objects for some environments can be harmful. As a result, there is a need to design a program that locates these differences using machine vision. The most important challenge of this problem is the change in lighting conditions and the presence of shadows in the scene. Therefore, the proposed methods must be resistant to these challenges. In this article, a method based on deep convolutional neural networks using transfer learning is introduced, which is trained with an intelligent data synthesis process. The results of this method are tested and presented on the dataset provided for this purpose. It is shown that the presented method is more efficient than other methods and can be used in a variety of real industrial environments.
translated by 谷歌翻译
Transformers have recently gained attention in the computer vision domain due to their ability to model long-range dependencies. However, the self-attention mechanism, which is the core part of the Transformer model, usually suffers from quadratic computational complexity with respect to the number of tokens. Many architectures attempt to reduce model complexity by limiting the self-attention mechanism to local regions or by redesigning the tokenization process. In this paper, we propose DAE-Former, a novel method that seeks to provide an alternative perspective by efficiently designing the self-attention mechanism. More specifically, we reformulate the self-attention mechanism to capture both spatial and channel relations across the whole feature dimension while staying computationally efficient. Furthermore, we redesign the skip connection path by including the cross-attention module to ensure the feature reusability and enhance the localization power. Our method outperforms state-of-the-art methods on multi-organ cardiac and skin lesion segmentation datasets without requiring pre-training weights. The code is publicly available at https://github.com/mindflow-institue/DAEFormer.
translated by 谷歌翻译
The stock market prediction has been a traditional yet complex problem researched within diverse research areas and application domains due to its non-linear, highly volatile and complex nature. Existing surveys on stock market prediction often focus on traditional machine learning methods instead of deep learning methods. Deep learning has dominated many domains, gained much success and popularity in recent years in stock market prediction. This motivates us to provide a structured and comprehensive overview of the research on stock market prediction focusing on deep learning techniques. We present four elaborated subtasks of stock market prediction and propose a novel taxonomy to summarize the state-of-the-art models based on deep neural networks from 2011 to 2022. In addition, we also provide detailed statistics on the datasets and evaluation metrics commonly used in the stock market. Finally, we highlight some open issues and point out several future directions by sharing some new perspectives on stock market prediction.
translated by 谷歌翻译
Existing training criteria in automatic speech recognition(ASR) permit the model to freely explore more than one time alignments between the feature and label sequences. In this paper, we use entropy to measure a model's uncertainty, i.e. how it chooses to distribute the probability mass over the set of allowed alignments. Furthermore, we evaluate the effect of entropy regularization in encouraging the model to distribute the probability mass only on a smaller subset of allowed alignments. Experiments show that entropy regularization enables a much simpler decoding method without sacrificing word error rate, and provides better time alignment quality.
translated by 谷歌翻译
Climate change has increased the intensity, frequency, and duration of extreme weather events and natural disasters across the world. While the increased data on natural disasters improves the scope of machine learning (ML) in this field, progress is relatively slow. One bottleneck is the lack of benchmark datasets that would allow ML researchers to quantify their progress against a standard metric. The objective of this short paper is to explore the state of benchmark datasets for ML tasks related to natural disasters, categorizing them according to the disaster management cycle. We compile a list of existing benchmark datasets introduced in the past five years. We propose a web platform - NADBenchmarks - where researchers can search for benchmark datasets for natural disasters, and we develop a preliminary version of such a platform using our compiled list. This paper is intended to aid researchers in finding benchmark datasets to train their ML models on, and provide general directions for topics where they can contribute new benchmark datasets.
translated by 谷歌翻译
Recent pre-trained language models have shown promising capabilities in generating fluent and realistic natural language text. However, generating multi-sentence text with global content planning has been a long-existing research question. Current approaches for controlled text generation can hardly address this issue, as they usually condition on single known control attributes. In this study, we propose a low-cost yet effective framework which explicitly models the global content plan of the generated text. Specifically, it optimizes the joint distribution of the natural language sequence and the global content plan in a plug-and-play manner. We conduct extensive experiments on the well-established Recipe1M+ benchmark. Both automatic and human evaluations verify that our model achieves the state-of-the-art performance on the task of recipe generation
translated by 谷歌翻译